📚 node [[mini batch|mini batch]]
Welcome! Nobody has contributed anything to 'mini batch|mini batch' yet. You can:
-
Write something in the document below!
- There is at least one public document in every node in the Agora. Whatever you write in it will be integrated and made available for the next visitor to read and edit.
- Write to the Agora from social media.
-
Sign up as a full Agora user.
- As a full user you will be able to contribute your personal notes and resources directly to this knowledge commons. Some setup required :)
⥅ related node [[mini batch]]
⥅ related node [[mini batch_stochastic_gradient_descent_(sgd)]]
⥅ node [[mini-batch]] pulled by Agora
📓
garden/KGBicheno/Artificial Intelligence/Introduction to AI/Week 3 - Introduction/Definitions/Mini-Batch.md by @KGBicheno
mini-batch
Go back to the [[AI Glossary]]
A small, randomly selected subset of the entire batch of examples run together in a single iteration of training or inference. The batch size of a mini-batch is usually between 10 and 1,000. It is much more efficient to calculate the loss on a mini-batch than on the full training data.
⥅ node [[mini-batch_stochastic_gradient_descent_(sgd)]] pulled by Agora
📓
garden/KGBicheno/Artificial Intelligence/Introduction to AI/Week 3 - Introduction/Definitions/Mini-Batch_Stochastic_Gradient_Descent_(Sgd).md by @KGBicheno
mini-batch stochastic gradient descent (SGD)
Go back to the [[AI Glossary]]
A gradient descent algorithm that uses mini-batches. In other words, mini-batch SGD estimates the gradient based on a small subset of the training data. Vanilla SGD uses a mini-batch of size 1.
📖 stoas
- public document at doc.anagora.org/mini-batch|mini-batch
- video call at meet.jit.si/mini-batch|mini-batch
🔎 full text search for 'mini batch|mini batch'